首页> 外文OA文献 >GPU Asynchronous Stochastic Gradient Descent to Speed Up Neural Network Training
【2h】

GPU Asynchronous Stochastic Gradient Descent to Speed Up Neural Network Training

机译:GpU异步随机梯度下降加速神经网络   训练

摘要

The ability to train large-scale neural networks has resulted instate-of-the-art performance in many areas of computer vision. These resultshave largely come from computational break throughs of two forms: modelparallelism, e.g. GPU accelerated training, which has seen quick adoption incomputer vision circles, and data parallelism, e.g. A-SGD, whose large scalehas been used mostly in industry. We report early experiments with a systemthat makes use of both model parallelism and data parallelism, we call GPUA-SGD. We show using GPU A-SGD it is possible to speed up training of largeconvolutional neural networks useful for computer vision. We believe GPU A-SGDwill make it possible to train larger networks on larger training sets in areasonable amount of time.
机译:训练大规模神经网络的能力已导致在计算机视觉的许多领域具有最先进的性能。这些结果主要来自两种形式的计算突破:模型并行性,例如GPU加速训练,已在计算机视觉界迅速采用,并且数据并行性例如A-SGD的大规模使用已广泛应用于工业中。我们报告了使用模型并行性和数据并行性的系统进行的早期实验,我们将其称为GPUA-SGD。我们展示了使用GPU A-SGD可以加快对计算机视觉有用的大型卷积神经网络的训练。我们相信GPU A-SGD将使在可能的时间内在更大的训练集上训练更大的网络成为可能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号